{ "cells": [ { "cell_type": "markdown", "metadata": {}, "source": [ "# task2:Updated情感分析\n", "\n", "在上一小节中我们已经学习了情感分析的基本工作流程,下面我们将学习如何改进优化模型:涉及到如何使用压缩填充序列、加载和使用预训练词向量、采用不同的优化器、选择不同的RNN体系结构(包括双向RNN、多层RNN)和正则化。\n", "\n", "本章主要内容如下:\n", "- 序列填充\n", "- 预训练词嵌入\n", "- LSTM\n", "- 双向 RNN\n", "- 多层 RNN\n", "- 正则化\n", "- 优化" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "## 2.1 准备数据\n", "\n", "首先设置seed,并将其分类训练、测试、验证集。\n", "\n", "在准备数据的时候需要注意到,由于 RNN 只能处理序列中的非 padded 元素(即非0数据),对于任何 padded 元素输出都是 0 。所以注意到我们在准备数据的时候将include_length设置为True,以获得句子的实际长度,后续需要使用。" ] }, { "cell_type": "code", "execution_count": 1, "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/home/ben/miniconda3/envs/pytorch17/lib/python3.8/site-packages/torchtext-0.9.0a0+c38fd42-py3.8-linux-x86_64.egg/torchtext/data/field.py:150: UserWarning: Field class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.\n", " warnings.warn('{} class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.'.format(self.__class__.__name__), UserWarning)\n", "/home/ben/miniconda3/envs/pytorch17/lib/python3.8/site-packages/torchtext-0.9.0a0+c38fd42-py3.8-linux-x86_64.egg/torchtext/data/field.py:150: UserWarning: LabelField class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.\n", " warnings.warn('{} class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.'.format(self.__class__.__name__), UserWarning)\n" ] } ], "source": [ "import torch\n", "from torchtext.legacy import data\n", "\n", "SEED = 1234\n", "\n", "torch.manual_seed(SEED)\n", "torch.backends.cudnn.deterministic = True\n", "\n", "TEXT = data.Field(tokenize = 'spacy',\n", " tokenizer_language = 'en_core_web_sm',\n", " include_lengths = True)\n", "\n", "LABEL = data.LabelField(dtype = torch.float)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "#### 加载 IMDb 数据集" ] }, { "cell_type": "code", "execution_count": 2, "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/home/ben/miniconda3/envs/pytorch17/lib/python3.8/site-packages/torchtext-0.9.0a0+c38fd42-py3.8-linux-x86_64.egg/torchtext/data/example.py:78: UserWarning: Example class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.\n", " warnings.warn('Example class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.', UserWarning)\n" ] } ], "source": [ "from torchtext.legacy import datasets\n", "\n", "train_data, test_data = datasets.IMDB.splits(TEXT, LABEL)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "#### 从训练集中选取部分做验证集" ] }, { "cell_type": "code", "execution_count": 3, "metadata": {}, "outputs": [], "source": [ "import random\n", "\n", "train_data, valid_data = train_data.split(random_state = random.seed(SEED))" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "## 2.2 词向量" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "接下来,使用预训练词向量进行初始化操作,其中获取这些词向量是通过指定参数传递给 build_vocab 得到的。\n", "\n", "在这里,我们选取GloVe词向量,GloVe的全称是:Global Vectors for Word Representation。[此处](https://nlp.stanford.edu/projects/glove/)有关于其有详细的介绍和大量资源。本教程将不介绍该词向量是如何具体得到的,仅简单描述下如何使用此词向量,这里我们使用的是 \"glove.6B.100d\" ,其中,6B表示词向量是在60亿规模的tokens上训练得到的,100d表示词向量是100维的(注意,这个词向量有800多兆)\n", "\n", "当然也可以选择[其他的词向量](https://github.com/pytorch/text/blob/master/torchtext/vocab.py#L113)。理论上,这些预训练词向量在词嵌入向量空间中的距离在一定程度上表征了词之间的语义关系,例如,“terrible”、“awful”、“dreadful” ,它们的词嵌入向量空间的距离会非常近。\n", "\n", "TEXT.build_vocab表示从预训练的词向量中,将当前训练数据中的词汇的词向量抽取出来,构成当前训练集的 Vocab(词汇表)。对于当前词向量语料库中没有出现的单词(记为UNK,unknown),通过高斯分布随机初始化(unk_init = torch.Tensor.normal_)。" ] }, { "cell_type": "code", "execution_count": 4, "metadata": {}, "outputs": [], "source": [ "MAX_VOCAB_SIZE = 25_000\n", "\n", "TEXT.build_vocab(train_data, \n", " max_size = MAX_VOCAB_SIZE, \n", " vectors = \"glove.6B.100d\", \n", " unk_init = torch.Tensor.normal_)\n", "\n", "LABEL.build_vocab(train_data)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "## 2.3 创建迭代器+选取GPU" ] }, { "cell_type": "code", "execution_count": 5, "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/home/ben/miniconda3/envs/pytorch17/lib/python3.8/site-packages/torchtext-0.9.0a0+c38fd42-py3.8-linux-x86_64.egg/torchtext/data/iterator.py:48: UserWarning: BucketIterator class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.\n", " warnings.warn('{} class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.'.format(self.__class__.__name__), UserWarning)\n" ] } ], "source": [ "BATCH_SIZE = 64\n", "\n", "# 根据当前环境选择是否调用GPU进行训练\n", "device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n", "\n", "# 创建数据迭代器\n", "train_iterator, valid_iterator, test_iterator = data.BucketIterator.splits(\n", " (train_data, valid_data, test_data), \n", " batch_size = BATCH_SIZE,\n", " sort_within_batch = True,\n", " device = device)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "## 2.4 构建模型\n", "\n", "\n", "### LSTM\n", "\n", "LSTM是标准RNN的一种变体,它增加了一种携带信息跨越多个时间步的方法,一定程度上克服了标准的RNN存在[梯度消失](https://en.wikipedia.org/wiki/Vanishing_gradient_problem)的问题。 具体来说,LSTM加入记忆单元$c_t$,其可看作是LSTM的“内存”,存储了时刻$t$时LSTM的记忆,可以认为其中保存了从过去到时刻$t$的所有必要信息,同时使用多个门控制信息流入和流出内存,具体地看[此处](https://colah.github.io/posts/2015-08-Understanding-LSTMs/)。 因此我们可以将LSTM看作是$x_t$、$h_t$和$c_t$的函数,而不仅仅是$x_t$和$h_t$。\n", "\n", "$$(h_t, c_t) = \\text{LSTM}(x_t, h_t, c_t)$$\n", "\n", "因此,使用LSTM的模型结构看起来类似下面这种构造(省略了嵌入层):\n", "\n", "![](assets/sentiment2.png)\n", "\n", "与初始隐藏状态一样,初始记忆状态$c_0$初始化为全零张量。需要注意的是,情感预测只使用最终隐藏状态,而不是最终记忆单元状态,即 $\\hat{y}=f(h_T)$。\n", "\n", "\n", "### 双向RNN\n", "\n", "\n", "双向RNN在之前的标准RNN层上添加了一个反方向处理的RNN层。然后,拼接各个时刻的两个RNN层的隐藏状态,将其作为最后的隐藏状态向量。也就是在时间步$t$时,前向RNN处理单词$x_t$,后向RNN处理单词$x_{T-t+1}$。通过这样的双向处理,每个单词对应的隐藏状态可以从左右两个方向聚集信息,这样一来,这些向量就编码了更均衡的信息。\n", "\n", "我们使用前向RNN的最后一个隐藏状态(从句子的最后一个单词获得)$h_T^\\rightarrow$和后向RNN的最后一个隐藏状态(从句子的第一个单词获得)$h_T^\\leftarrow$进行情感预测,即 $\\hat{y}=f(h_T^\\rightarrow, h_T^\\leftarrow)$,下图显示了一个双向RNN,前向RNN为橙色,后向RNN为绿色,线性层为银色。\n", "\n", "![](assets/sentiment3.png)\n", "\n", "\n", "\n", "### 多层RNN\n", "\n", "多层RNN(也称为深层RNN):在初始标准RNN上多加几层RNN。第一个(底部)RNN在时间步$t$时输出的隐藏状态将是在时间步$t$时其上方RNN的输入,然后根据最终(最高)层的最终隐藏状态进行预测。\n", "下图显示了多层单向RNN,其中层号以上标形式给出。还要注意,每个层都需要自己的初始隐藏状态,$h_0^L$。\n", "\n", "![](assets/sentiment4.png)\n", "\n", "\n", "### 正则化\n", "\n", "我们已经对模型进行了各方面的改进,但是我们需要注意的是,当模型的参数逐渐增加的同时,模型过拟合的可能性就越大。为了解决这个问题,我们添加dropout正则化。Dropout的工作原理是在前向传播过程中,层中的神经元随机*Dropout*(设置为0)。每个神经元是否被drop的概率则由一个超参数设置,并不受其他神经元影响。\n", "\n", "关于为什么dropout有效的一种理论是,参数dropout的模型可以被视为“weaker”(参数较少)的模型。因此,最终的模型可以被认为是所有这些weaker模型的集合,这些模型都没有过度参数化,因此降低了过拟合的可能性。\n", "\n", "### Implementation Details\n", "\n", "1.针对模型训练过程中的一点补充:在模型训练过程中,对于每个样本中补齐后加上的pad token,模型是不应该对其进行训练的,也就是并不会学习“\\”标记的嵌入。因为padding token跟句子的情感是无关的。这就意味着pad token的嵌入层(词向量)会一直保持初始化的状态(初始化为全零)。具体而言,我们是通过往nn.Embedding 层传入pad token 的index索引,作为padding_idx参数。\n", "\n", "2.因为实验中使用的双向LSTM的包含了前向传播和后向传播过程,所以最后的隐藏状态向量包含了前向和后向的隐藏状态,所以在下一层nn.Linear层中的输入的形状就是隐藏层维度形状的两倍。\n", " \n", "3.在将embeddings(词向量)输入RNN前,我们需要借助`nn.utils.rnn.packed_padded_sequence`将它们‘打包’,以此来保证RNN只会处理不是pad的token。我们得到的输出包括`packed_output` (a packed sequence)以及`hidden sate` 和 `cell state`。如果没有进行‘打包’操作,那么输出的`hidden state`和`cell state`大概率是来自句子的pad token。如果使用packed padded sentences,输出的就会是最后一个非padded元素的`hidden state` 和 `cell state`。\n", "\n", "4.之后我们借助`nn.utils.rnn.pad_packed_sequence` 将输出的句子‘解压’转换成一个tensor张量。需要注意的是来自padding tokens的输出是零张量,通常情况下,我们只有在后续的模型中使用输出时才需要‘解压’。虽然在本案例中下不需要,这里只是为展示其步骤。\n", "\n", "5.final hidden sate:也就是hidden,其形状是[num layers * num directions, batch size, hid dim]。因为我们只要最后的前向和后向传播的hidden states,我们只要最后2个hidden layers就行hidden[-2,:,:] 和hidden[-1,:,:],然后将他们合并在一起,再传入线性层linear layer。 \n", "#####这里不知道怎么解释会比较好,还需调整。\n" ] }, { "cell_type": "code", "execution_count": 6, "metadata": {}, "outputs": [], "source": [ "import torch.nn as nn\n", "\n", "class RNN(nn.Module):\n", " def __init__(self, vocab_size, embedding_dim, hidden_dim, output_dim, n_layers, \n", " bidirectional, dropout, pad_idx):\n", " \n", " super().__init__()\n", " # embedding嵌入层(词向量)\n", " self.embedding = nn.Embedding(vocab_size, embedding_dim, padding_idx = pad_idx)\n", " \n", " # RNN变体——双向LSTM\n", " self.rnn = nn.LSTM(embedding_dim, # input_size\n", " hidden_dim, #output_size\n", " num_layers=n_layers, # 层数\n", " bidirectional=bidirectional, #是否双向\n", " dropout=dropout) #随机去除神经元\n", " # 线性连接层\n", " self.fc = nn.Linear(hidden_dim * 2, output_dim) # 因为前向传播+后向传播有两个hidden sate,且合并在一起,所以乘以2\n", " \n", " # 随机去除神经元\n", " self.dropout = nn.Dropout(dropout)\n", " \n", " def forward(self, text, text_lengths):\n", " \n", " #text 的形状 [sent len, batch size]\n", " \n", " embedded = self.dropout(self.embedding(text))\n", " \n", " #embedded 的形状 [sent len, batch size, emb dim]\n", " \n", " # pack sequence\n", " # lengths need to be on CPU!\n", " packed_embedded = nn.utils.rnn.pack_padded_sequence(embedded, text_lengths.to('cpu'))\n", " \n", " packed_output, (hidden, cell) = self.rnn(packed_embedded)\n", " \n", " #unpack sequence\n", " output, output_lengths = nn.utils.rnn.pad_packed_sequence(packed_output)\n", "\n", " #output的形状[sent len, batch size, hid dim * num directions]\n", " #output中的 padding tokens是数值为0的张量\n", " \n", " #hidden 的形状 [num layers * num directions, batch size, hid dim]\n", " #cell 的形状 [num layers * num directions, batch size, hid dim]\n", " \n", " #concat the final forward (hidden[-2,:,:]) and backward (hidden[-1,:,:]) hidden layers\n", " #and apply dropout\n", " \n", " hidden = self.dropout(torch.cat((hidden[-2,:,:], hidden[-1,:,:]), dim = 1))\n", " \n", " #hidden 的形状 [batch size, hid dim * num directions]\n", " \n", " return self.fc(hidden)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "## 2.5 实例化模型+传入参数\n", "\n", "为了保证pre-trained 词向量可以加载到模型中,EMBEDDING_DIM 必须等于预训练的GloVe词向量的大小。" ] }, { "cell_type": "code", "execution_count": 7, "metadata": {}, "outputs": [], "source": [ "INPUT_DIM = len(TEXT.vocab) # 250002: 之前设置的只取25000个最频繁的词,加上pad_token和unknown token\n", "EMBEDDING_DIM = 100\n", "HIDDEN_DIM = 256\n", "OUTPUT_DIM = 1\n", "N_LAYERS = 2\n", "BIDIRECTIONAL = True\n", "DROPOUT = 0.5\n", "PAD_IDX = TEXT.vocab.stoi[TEXT.pad_token] #指定参数,定义pad_token的index索引值,让模型不管pad token\n", "\n", "model = RNN(INPUT_DIM, \n", " EMBEDDING_DIM, \n", " HIDDEN_DIM, \n", " OUTPUT_DIM, \n", " N_LAYERS, \n", " BIDIRECTIONAL, \n", " DROPOUT, \n", " PAD_IDX)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "查看模型参数量" ] }, { "cell_type": "code", "execution_count": 8, "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "The model has 4,810,857 trainable parameters\n" ] } ], "source": [ "def count_parameters(model):\n", " return sum(p.numel() for p in model.parameters() if p.requires_grad)\n", "\n", "print(f'The model has {count_parameters(model):,} trainable parameters')" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "接下来,把前面加载好的预训练词向量复制进我们模型中的embedding嵌入层,用预训练的embeddings词向量替换掉原来模型初始化的权重参数。\n", "\n", "我们从字段的vocab中检索嵌入,并检查它们的大小是否正确,_**[vocab size, embedding dim]**_ " ] }, { "cell_type": "code", "execution_count": 9, "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "torch.Size([25002, 100])\n" ] } ], "source": [ "pretrained_embeddings = TEXT.vocab.vectors\n", "# 检查词向量形状 [vocab size, embedding dim]\n", "print(pretrained_embeddings.shape)" ] }, { "cell_type": "code", "execution_count": 10, "metadata": {}, "outputs": [ { "data": { "text/plain": [ "tensor([[-0.1117, -0.4966, 0.1631, ..., 1.2647, -0.2753, -0.1325],\n", " [-0.8555, -0.7208, 1.3755, ..., 0.0825, -1.1314, 0.3997],\n", " [-0.0382, -0.2449, 0.7281, ..., -0.1459, 0.8278, 0.2706],\n", " ...,\n", " [ 0.6783, 0.0488, 0.5860, ..., 0.2680, -0.0086, 0.5758],\n", " [-0.6208, -0.0480, -0.1046, ..., 0.3718, 0.1225, 0.1061],\n", " [-0.6553, -0.6292, 0.9967, ..., 0.2278, -0.1975, 0.0857]])" ] }, "execution_count": 10, "metadata": {}, "output_type": "execute_result" } ], "source": [ "# 用预训练的embedding词向量替换原始模型初始化的权重参数\n", "model.embedding.weight.data.copy_(pretrained_embeddings)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "因为我们的`` 和`` token不在预训练词表里,它们已经在构建我们自己的词表时,使用 `unk_init` (an $\\mathcal{N}(0,1)$ distribution)初始化了。所以,最好显式地告诉模型,将它们初始化变为0,它们与情感无关。\n", " \n", "我们是通过手动设置他们的词向量权重为0的。\n", " \n", "**注意**:与初始化嵌入一样,这应该在“weight.data”而不是“weight”上完成!" ] }, { "cell_type": "code", "execution_count": 11, "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "tensor([[ 0.0000, 0.0000, 0.0000, ..., 0.0000, 0.0000, 0.0000],\n", " [ 0.0000, 0.0000, 0.0000, ..., 0.0000, 0.0000, 0.0000],\n", " [-0.0382, -0.2449, 0.7281, ..., -0.1459, 0.8278, 0.2706],\n", " ...,\n", " [ 0.6783, 0.0488, 0.5860, ..., 0.2680, -0.0086, 0.5758],\n", " [-0.6208, -0.0480, -0.1046, ..., 0.3718, 0.1225, 0.1061],\n", " [-0.6553, -0.6292, 0.9967, ..., 0.2278, -0.1975, 0.0857]])\n" ] } ], "source": [ "#将unknown 和padding token设置为0\n", "UNK_IDX = TEXT.vocab.stoi[TEXT.unk_token]\n", "\n", "model.embedding.weight.data[UNK_IDX] = torch.zeros(EMBEDDING_DIM)\n", "model.embedding.weight.data[PAD_IDX] = torch.zeros(EMBEDDING_DIM)\n", "\n", "print(model.embedding.weight.data)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "我们现在可以看到嵌入权重矩阵的前两行值都是0,需要注意的是,pad token的词向量在模型训练过程中始终不会被学习。而unknown token的词向量是会被学习的。" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "## 2.6 训练模型\n", "\n", "现在开始训练模型!\n", "\n", "我们是将随机梯度下降优化器从'SGD'更改为'Adam'。SGD使用我们设定好的学习率同步更新所有参数,而Adam会调整每个参数的学习率,给出更新频率更高的参数,以及更新频率更低的参数和更新频率不高的参数。有关“Adam”(和其他优化器)的更多信息,请参见[此处](http://ruder.io/optimizing-gradient-descent/index.html)。\n", " \n", "要将'SGD'更改为'Adam',我们只需将'optim.SGD'更改为'optim.Adam',还要注意,我们不提供 Adam初始学习率,因为PyTorch提供了默认的初始学习率。\n", "\n", "### 2.6.1 设置优化器" ] }, { "cell_type": "code", "execution_count": 12, "metadata": {}, "outputs": [], "source": [ "import torch.optim as optim\n", "\n", "optimizer = optim.Adam(model.parameters())" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "### 2.6.2 设置损失函数和GPU\n", "\n", "训练模型的其它步骤保持不变。" ] }, { "cell_type": "code", "execution_count": 13, "metadata": {}, "outputs": [], "source": [ "criterion = nn.BCEWithLogitsLoss() # 损失函数. criterion 在本task中时损失函数的意思\n", "\n", "model = model.to(device)\n", "criterion = criterion.to(device)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "### 2.6.3 计算精确度" ] }, { "cell_type": "code", "execution_count": 14, "metadata": {}, "outputs": [], "source": [ "def binary_accuracy(preds, y):\n", " \"\"\"\n", " Returns accuracy per batch, i.e. if you get 8/10 right, this returns 0.8, NOT 8\n", " \"\"\"\n", "\n", " #round predictions to the closest integer\n", " rounded_preds = torch.round(torch.sigmoid(preds))\n", " correct = (rounded_preds == y).float() #convert into float for division \n", " acc = correct.sum() / len(correct)\n", " return acc" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "### 2.6.4 定义一个训练函数,用来训练模型\n", "\n", "正如我们设置的“include_length=True”,我们的“batch.text”现在是一个元组,第一个元素是数字张量,第二个元素是每个序列的实际长度。在将它们传递给模型之前,我们将它们分成各自的变量“text”和“text_length”。\n", "\n", "**注意**:因为现在使用的是dropout,我们必须记住使用`model.train()`以确保在训练时开启 dropout。\n" ] }, { "cell_type": "code", "execution_count": 15, "metadata": {}, "outputs": [], "source": [ "def train(model, iterator, optimizer, criterion):\n", " \n", " epoch_loss = 0\n", " epoch_acc = 0\n", " \n", " model.train()\n", " \n", " for batch in iterator:\n", " \n", " optimizer.zero_grad() # 梯度清零\n", " \n", " text, text_lengths = batch.text # batch.text返回的是一个元组(数字化的张量,每个句子的长度) \n", " \n", " predictions = model(text, text_lengths).squeeze(1)\n", " \n", " loss = criterion(predictions, batch.label)\n", " \n", " acc = binary_accuracy(predictions, batch.label)\n", " \n", " loss.backward()\n", " \n", " optimizer.step()\n", " \n", " epoch_loss += loss.item()\n", " epoch_acc += acc.item()\n", " \n", " return epoch_loss / len(iterator), epoch_acc / len(iterator)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "### 2.6.5 定义一个测试函数\n", "\n", "**注意**:因为现在使用的是dropout,我们必须记住使用`model.eval()`以确保在评估时关闭 dropout。" ] }, { "cell_type": "code", "execution_count": 16, "metadata": {}, "outputs": [], "source": [ "def evaluate(model, iterator, criterion):\n", " \n", " epoch_loss = 0\n", " epoch_acc = 0\n", " \n", " model.eval()\n", " \n", " with torch.no_grad():\n", " \n", " for batch in iterator:\n", "\n", " text, text_lengths = batch.text #batch.text返回的是一个元组(数字化的张量,每个句子的长度) \n", " \n", " predictions = model(text, text_lengths).squeeze(1)\n", " \n", " loss = criterion(predictions, batch.label)\n", " \n", " acc = binary_accuracy(predictions, batch.label)\n", "\n", " epoch_loss += loss.item()\n", " epoch_acc += acc.item()\n", " \n", " return epoch_loss / len(iterator), epoch_acc / len(iterator)" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "还可以创建一个函数告诉我们epochs训练需要多长时间。" ] }, { "cell_type": "code", "execution_count": 17, "metadata": {}, "outputs": [], "source": [ "import time\n", "\n", "def epoch_time(start_time, end_time):\n", " elapsed_time = end_time - start_time\n", " elapsed_mins = int(elapsed_time / 60)\n", " elapsed_secs = int(elapsed_time - (elapsed_mins * 60))\n", " return elapsed_mins, elapsed_secs" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "### 2.6.6 正式训练模型" ] }, { "cell_type": "code", "execution_count": 18, "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/home/ben/miniconda3/envs/pytorch17/lib/python3.8/site-packages/torchtext-0.9.0a0+c38fd42-py3.8-linux-x86_64.egg/torchtext/data/batch.py:23: UserWarning: Batch class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.\n", " warnings.warn('{} class will be retired soon and moved to torchtext.legacy. Please see the most recent release notes for further information.'.format(self.__class__.__name__), UserWarning)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch: 01 | Epoch Time: 0m 36s\n", "\tTrain Loss: 0.673 | Train Acc: 58.05%\n", "\t Val. Loss: 0.619 | Val. Acc: 64.97%\n", "Epoch: 02 | Epoch Time: 0m 36s\n", "\tTrain Loss: 0.611 | Train Acc: 66.33%\n", "\t Val. Loss: 0.510 | Val. Acc: 74.32%\n", "Epoch: 03 | Epoch Time: 0m 37s\n", "\tTrain Loss: 0.484 | Train Acc: 77.04%\n", "\t Val. Loss: 0.397 | Val. Acc: 82.95%\n", "Epoch: 04 | Epoch Time: 0m 37s\n", "\tTrain Loss: 0.384 | Train Acc: 83.57%\n", "\t Val. Loss: 0.407 | Val. Acc: 83.23%\n", "Epoch: 05 | Epoch Time: 0m 37s\n", "\tTrain Loss: 0.314 | Train Acc: 86.98%\n", "\t Val. Loss: 0.314 | Val. Acc: 86.36%\n" ] } ], "source": [ "N_EPOCHS = 5\n", "\n", "best_valid_loss = float('inf')\n", "\n", "for epoch in range(N_EPOCHS):\n", "\n", " start_time = time.time()\n", " \n", " train_loss, train_acc = train(model, train_iterator, optimizer, criterion)\n", " valid_loss, valid_acc = evaluate(model, valid_iterator, criterion)\n", " \n", " end_time = time.time()\n", "\n", " epoch_mins, epoch_secs = epoch_time(start_time, end_time)\n", " # 保留最好的训练结果的那个模型参数,之后加载这个进行预测\n", " if valid_loss < best_valid_loss:\n", " best_valid_loss = valid_loss\n", " torch.save(model.state_dict(), 'tut2-model.pt')\n", " \n", " print(f'Epoch: {epoch+1:02} | Epoch Time: {epoch_mins}m {epoch_secs}s')\n", " print(f'\\tTrain Loss: {train_loss:.3f} | Train Acc: {train_acc*100:.2f}%')\n", " print(f'\\t Val. Loss: {valid_loss:.3f} | Val. Acc: {valid_acc*100:.2f}%')" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "### 2.6.7 最终测试结果" ] }, { "cell_type": "code", "execution_count": 19, "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "Test Loss: 0.334 | Test Acc: 85.28%\n" ] } ], "source": [ "model.load_state_dict(torch.load('tut2-model.pt'))\n", "\n", "test_loss, test_acc = evaluate(model, test_iterator, criterion)\n", "\n", "print(f'Test Loss: {test_loss:.3f} | Test Acc: {test_acc*100:.2f}%')" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "## 2.7 模型验证\n", "\n", "我们现在可以使用这个模型来预测我们给出的任何句子的情感了,注意需要提供的句子是电影评论方面的。\n", "\n", "当使用模型进行实际预测时,模型要始终在evaluation mode评估模式。\n", " \n", "“predict_sentiment”函数的作用如下:\n", "\n", "- 将模型切换为evaluate模式\n", "- 对句子进行分词操作\n", "- 将分词后的每个词,对应着词汇表,转换成对应的index索引,\n", "- 获取句子的长度\n", "- 将indexes,从list转化成tensor\n", "- 通过unsqueezing 添加一个batch维度\n", "- 将length转化成张量tensor\n", "- 用sigmoid函数将预测值压缩到0-1之间\n", "- 用item()方法,将只有一个值的张量tensor转化成整数\n", "\n", "负面评论返回接近0的值,正面评论返回接近1的值。" ] }, { "cell_type": "code", "execution_count": 20, "metadata": {}, "outputs": [], "source": [ "import spacy\n", "nlp = spacy.load('en_core_web_sm')\n", "\n", "def predict_sentiment(model, sentence):\n", " model.eval()\n", " tokenized = [tok.text for tok in nlp.tokenizer(sentence)]\n", " indexed = [TEXT.vocab.stoi[t] for t in tokenized]\n", " length = [len(indexed)]\n", " tensor = torch.LongTensor(indexed).to(device)\n", " tensor = tensor.unsqueeze(1)\n", " length_tensor = torch.LongTensor(length)\n", " prediction = torch.sigmoid(model(tensor, length_tensor))\n", " return prediction.item()" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "负面评论的例子:" ] }, { "cell_type": "code", "execution_count": 21, "metadata": {}, "outputs": [ { "data": { "text/plain": [ "0.05380420759320259" ] }, "execution_count": 21, "metadata": {}, "output_type": "execute_result" } ], "source": [ "predict_sentiment(model, \"This film is terrible\")" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "正面评论的例子:" ] }, { "cell_type": "code", "execution_count": 22, "metadata": {}, "outputs": [ { "data": { "text/plain": [ "0.94941645860672" ] }, "execution_count": 22, "metadata": {}, "output_type": "execute_result" } ], "source": [ "predict_sentiment(model, \"This film is great\")" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "## 小结\n", "我们现在已经为电影评论建立了一个情感分析模型。在下一小节中,我们将实现一个模型,这个模型会以更少的参数获得更高的精度、更快的训练速度。\n", "\n", "\n" ] } ], "metadata": { "kernelspec": { "display_name": "Python 3", "language": "python", "name": "python3" }, "language_info": { "codemirror_mode": { "name": "ipython", "version": 3 }, "file_extension": ".py", "mimetype": "text/x-python", "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", "version": "3.7.10" } }, "nbformat": 4, "nbformat_minor": 4 }